Interpolation of stochastic grammar and word bigram models in natural language understanding

نویسندگان

  • Sven C. Martin
  • Andreas Kellner
  • Thomas Portele
چکیده

The paper shows the effects of combining a stochastic grammar with a word bigram language model by log-linear interpolation. It is divided into three main parts: The first part derives the stochastic grammar model and gives a sound theoretical motivation to incorporate word dependencies such as bigrams. The second part describes two different algorithmic approaches to the combination of both models by log-linear interpolation. The third part reports attribute error rate (AER) results measured on the Philips corpus of train time table inquiries that show a reduction of up to 9% relative. 1. STOCHASTIC MODEL OF NATURAL LANGUAGE UNDERSTANDING The Philips Natural Language Understanding (NLU) module is used in automated inquiry systems (AIS), such as train table enquiries [2], to analyze the word sequence of a user utterance. It does not try to find parse trees that cover the whole word sequence but breaks up the sequence into chunks, where each chunk belongs to a semantically meaningful concept. A stochastic context–free grammar is used to derive the word chunk from a concept. The chunking is useful since the spontaneous speech that occurs in dialogue applications is very ungrammatical. Thus, a robust NLU model concentrates on the useful parts of a user utterance. Other recent works also employ some kind of chunking, e.g. [6, 9]. The stochastic model of the Philips NLU module was developed by H. Aust in [1, p. 81]. Here, we show that this model can be derived from Bayes’ decision rule. This derivation gives a sound theoretical motivation to incorporate word dependencies such as bigrams. Bayes’ decision rule finds the most likely concept sequence K̂ = k̂1, . . . , k̂s, given the sequence O = o1, . . . , ot of acoustic observations. The derivation of the concept sequence K does not directly depend on the acoustic observations O but on a word sequence W = w1, . . . , wN derived from O as an intermediate result: K̂ = argmax K p(K|O)

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Permugram language models

In natural languages, the words within an utterance are often correlated over large distances. Long-spanning con-textual eeects of this type cannot be eeciently and robustly captured by the traditional N-gram approaches of stochastic language modelling. We present a new kind of stochastic grammar | the permugram model. A permu-gram model is obtained by linear interpolation of a large number of ...

متن کامل

Experience with a Stack Decoder-Based HMM CSR and Back-Off N-Gram Language Models

Stochastic language models are more useful than nonstochastic models because they contribute more information than a simple acceptance or rejection of a word sequence. Back-off N-gram language models [ I l l are an effective class of word based stochastic language model. The first part of this paper describes our experiences using the back-off language models in our time-synchronous decoder CSR...

متن کامل

Using a stochastic context-free grammar as a language model for speech recognition

This paper describes a number of experiments in adding new grammatical knowledge to the Berkeley Restaurant Project (BeRP), our medium-vocabulary (1300 word), speaker-independent, spontaneous continuous-speech understanding system (Jurafsky et al. 1994). We describe an algorithm for using a probabilistic Earley parser and a stochastic context-free grammar (SCFG) to generate word transition prob...

متن کامل

Parsing N-Best Lists of Handwritten Sentences

This paper investigates the application of a probabilistic parser for natural language on the list of the Nbest sentences produced by an off-line recognition system for cursive handwritten sentences. For the generation of the N-best sentence list an HMM-based recognizer including a bigram language model is used. The parsing of the sentences is achieved by a bottom-up chart parser for stochastic...

متن کامل

A word graph interface for a flexible concept based speech understanding framework

In this paper, we introduce a word graph interface between speech and natural language processing systems within a flexible speech understanding framework based on stochastic concept modeling augmented with background ”filler” models. Each concept represents a set of phrases ( written as a context free grammar (CFG)) with the same meaning, and is compiled into a stochastic recursive transition ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2000